26 research outputs found
An efficient RANSAC hypothesis evaluation using sufficient statistics for RGB-D pose estimation
Achieving autonomous flight in GPS-denied environments begins with pose estimation in three-dimensional space, and this is much more challenging in an MAV in a swarm robotic system due to limited computational resources. In vision-based pose estimation, outlier detection is the most time-consuming step. This usually involves a RANSAC procedure using the reprojection-error method for hypothesis evaluation. Realignment-based hypothesis evaluation method is observed to be more accurate, but the considerably slower speed makes it unsuitable for robots with limited resources. We use sufficient statistics of least-squares minimisation to speed up this process. The additive nature of these sufficient statistics makes it possible to compute pose estimates in each evaluation by reusing previously computed statistics. Thus estimates need not be calculated from scratch each time. The proposed method is tested on standard RANSAC, Preemptive RANSAC and R-RANSAC using benchmark datasets. The results show that the use of sufficient statistics speeds up the outlier detection process with realignment hypothesis evaluation for all RANSAC variants, achieving an execution speed of up to 6.72 times
Reinforcement Learning for Shared Autonomy Drone Landings
Novice pilots find it difficult to operate and land unmanned aerial vehicles
(UAVs), due to the complex UAV dynamics, challenges in depth perception, lack
of expertise with the control interface and additional disturbances from the
ground effect. Therefore we propose a shared autonomy approach to assist pilots
in safely landing a UAV under conditions where depth perception is difficult
and safe landing zones are limited. Our approach comprises of two modules: a
perception module that encodes information onto a compressed latent
representation using two RGB-D cameras and a policy module that is trained with
the reinforcement learning algorithm TD3 to discern the pilot's intent and to
provide control inputs that augment the user's input to safely land the UAV.
The policy module is trained in simulation using a population of simulated
users. Simulated users are sampled from a parametric model with four
parameters, which model a pilot's tendency to conform to the assistant,
proficiency, aggressiveness and speed. We conduct a user study (n = 28) where
human participants were tasked with landing a physical UAV on one of several
platforms under challenging viewing conditions. The assistant, trained with
only simulated user data, improved task success rate from 51.4% to 98.2%
despite being unaware of the human participants' goal or the structure of the
environment a priori. With the proposed assistant, regardless of prior piloting
experience, participants performed with a proficiency greater than the most
experienced unassisted participants.Comment: 14 pages, 13 figures. Submitted to IEEE Transactions on Robotics
(T-RO